DeepSeek Affiliate Company Releases Innovative Deployment Patent for Large Language Models, Promoting New Developments in AI Technology
DeepSeek affiliate company publicly disclosed a patent for deploying large language models, innovatively utilizing a distributed architecture: deploying the prefill and decode stages on high-performance computing and large-memory machines respectively. This method balances the load, reduces computational idle time, significantly lowering latency and increasing throughput. The patent highlights optimizations in system scalability and fault tolerance. Its MoE language model, DeepSeek-V3, has 671 billion parameters, with 37 billion parameters activated per token, which will promote the practical application of AI technology across industries. The core breakthrough lies in combining hardware resources effectively.